28 research outputs found
Recommended from our members
Practicalities of predicting quality of high dynamic range images and video
© 2016 IEEE.The paper discusses the use of existing metrics, such as HDR-VDP and extensions of MS-SSIM and PSNR, for prediction of quality in high dynamic range (HDR) images and video. The discussion is based on the experience in using those metrics to evaluate and improve image compression for the new JPEG XT standard, and video compression for the LumaHDR open source codec. The paper explains why existing non-HDR metrics perform very poorly on HDR data and how to improve their predictions. Since most HDR metrics require calibrated data, intended for an HDR display, such calibration step is explained. One of the popular HDR quality metrics, HDR-VDP, is briefly introduced with the update on the latest improvements. Finally, several studies comparing objective HDR metric performance are summarized
Exploiting the limitations of spatio-temporal vision for more efficient VR rendering
Increasingly higher virtual reality (VR) display resolutions and
good-quality anti-aliasing make rendering in VR prohibitively expensive.
The generation of these complex frames 90 times per second
in a binocular setup demands substantial computational power.
Wireless transmission of the frames from the GPU to the VR headset
poses another challenge, requiring high-bandwidth dedicated
links
Recommended from our members
The effect of display brightness and viewing distance: A dataset for visually lossless image compression
Visibility of image artifacts depends on the viewing conditions, such as display brightness and distance to the display. However, most image and video quality metrics operate under the assumption of a single standard viewing condition, without considering luminance or distance to
the display. To address this limitation, we isolate brightness and distance as the components impacting the visibility of artifacts and collect a new dataset for visually lossless image compression. The dataset includes images encoded with JPEG andWebP at the quality level that makes compression
artifacts imperceptible to an average observer. The visibility thresholds are collected under two luminance conditions: 10 cd/m2, simulating a dimmed mobile phone, and 220 cd/m2, which is a typical peak luminance of modern computer displays; and two distance conditions:
30 and 60 pixels per visual degree. The dataset was used to evaluate existing image quality and visibility metrics in their ability to consider display brightness and its distance to viewer. We include two deep neural network architectures, proposed to control image compression for visually
lossless coding in our experiments.
</jats:p
Noise-Aware Merging of High Dynamic Range Image Stacks Without Camera Calibration
A near-optimal reconstruction of the radiance of a High Dynamic Range scene
from an exposure stack can be obtained by modeling the camera noise
distribution. The latent radiance is then estimated using Maximum Likelihood
Estimation. But this requires a well-calibrated noise model of the camera,
which is difficult to obtain in practice. We show that an unbiased estimation
of comparable variance can be obtained with a simpler Poisson noise estimator,
which does not require the knowledge of camera-specific noise parameters. We
demonstrate this empirically for four different cameras, ranging from a
smartphone camera to a full-frame mirrorless camera. Our experimental results
are consistent for simulated as well as real images, and across different
camera settings
Trained Perceptual Transform for Quality Assessment of High Dynamic Range Images and Video
In this paper, we propose a trained perceptually transform for
quality assessment of high dynamic range (HDR) images and
video. The transform is used to convert absolute luminance
values found in HDR images into perceptually uniform units,
which can be used with any standard-dynamic-range metric.
The new transform is derived by fitting the parameters
of a previously proposed perceptual encoding function to 4
different HDR subjective quality assessment datasets using
Bayesian optimization. The new transform combined with a
simple peak signal-to-noise ratio measure achieves better prediction
performance in cross-dataset validation than existing
transforms. We provide Matlab code for our metri
Recommended from our members
Perceptual image quality assessment for various viewing conditions and display systems
From complete darkness to direct sunlight, real-world dis-
plays operate in various viewing conditions often resulting in a
non-optimal viewing experience. Most existing Image Quality
Assessment (IQA) methods, however, assume ideal environments
and displays, and thus cannot be used when viewing conditions
differ from the standard. In this paper, we investigate the influence
of ambient illumination level and display luminance on human
perception of image quality. We conduct a psychophysical study
to collect a novel dataset of over 10000 image quality preference
judgments performed in illumination conditions ranging from 0 lux
to 20000 lux. We also propose a perceptual IQA framework that
allows most existing image quality metrics (IQM) to accurately
predict image quality for a wide range of illumination conditions
and display parameters 1 . Our analysis demonstrates strong cor-
relation between human IQA and the predictions of our proposed
framework combined with multiple prominent IQMs and across a
wide range of luminance values
Dataset and metrics for predicting local visible differences
A large number of imaging and computer graphics applications require localized information on the visibility of image distortions. Existing image quality metrics are not suitable for this task as they provide a single quality value per image. Existing visibility metrics produce visual difference maps, and are specifically designed for detecting just noticeable distortions but their predictions are often inaccurate. In this work, we argue that the key reason for this problem is the lack of large image collections with a good coverage of possible distortions that occur in different applications. To address the problem, we collect an extensive dataset of reference and distorted image pairs together with user markings indicating whether distortions are visible or not. We propose a statistical model that is designed for the meaningful interpretation of such data, which is affected by visual search and imprecision of manual marking. We use our dataset for training existing metrics and we demonstrate that their performance significantly improves. We show that our dataset with the proposed statistical model can be used to train a new CNN-based metric, which outperforms the existing solutions. We demonstrate the utility of such a metric in visually lossless JPEG compression, super-resolution and watermarking.</jats:p
Real-time noise-aware tone mapping
Real-time high quality video tone mapping is needed for many applications, such as digital viewfinders in cameras, display algorithms which adapt to ambient light, in-camera processing, rendering engines for video games and video post-processing. We propose a viable solution for these applications by designing a video tone-mapping operator that controls the visibility of the noise, adapts to display and viewing environment, minimizes contrast distortions, preserves or enhances image details, and can be run in real-time on an incoming sequence without any preprocessing. To our knowledge, no existing solution offers all these features. Our novel contributions are: a fast procedure for computing local display-adaptive tone-curves which minimize contrast distortions, a fast method for detail enhancement free from ringing artifacts, and an integrated video tone-mapping solution combining all the above features.This project was funded by the Swedish Foundation for Strategic Research (SSF) through grant IIS11-0081, Linkoping University Center for Industrial Information Technology (CENIIT), the Swedish Research Council through the Linnaeus Environment CADICS, and through COST Action IC1005
Hybrid-MST: A hybrid active sampling strategy for pairwise preference aggregation
In this paper we present a hybrid active sampling strategy for pairwise
preference aggregation, which aims at recovering the underlying rating of the
test candidates from sparse and noisy pairwise labelling. Our method employs
Bayesian optimization framework and Bradley-Terry model to construct the
utility function, then to obtain the Expected Information Gain (EIG) of each
pair. For computational efficiency, Gaussian-Hermite quadrature is used for
estimation of EIG. In this work, a hybrid active sampling strategy is proposed,
either using Global Maximum (GM) EIG sampling or Minimum Spanning Tree (MST)
sampling in each trial, which is determined by the test budget. The proposed
method has been validated on both simulated and real-world datasets, where it
shows higher preference aggregation ability than the state-of-the-art methods
Recommended from our members
Reproducing reality with a high-dynamic-range multi-focal stereo display
With well-established methods for producing photo-realistic results, the next big challenge of graphics and display technologies is to achieve perceptual realism --- producing imagery indistinguishable from real-world 3D scenes. To deliver all necessary visual cues for perceptual realism, we built a High-Dynamic-Range Multi-Focal Stereo Display that achieves high resolution, accurate color, a wide dynamic range, and most depth cues, including binocular presentation and a range of focal depth. The display and associated imaging system have been designed to capture and reproduce a small near-eye three-dimensional object and to allow for a direct comparison between virtual and real scenes. To assess our reproduction of realism and demonstrate the capability of the display and imaging system, we conducted an experiment in which the participants were asked to discriminate between a virtual object and its physical counterpart. Our results indicate that the participants can only detect the discrepancy with a probability of 0.44. With such a level of perceptual realism, our display apparatus can facilitate a range of visual experiments that require the highest fidelity of reproduction while allowing for the full control of the displayed stimuli.</jats:p